Comparison of PCA with ICA from data distribution perspective
نویسنده
چکیده
ICA algorithm has a similar procedure to PCA in terms of minimizing the objective function and deriving optimal weights. We used iterative version of PCA and ICA exactly to be able to notice that fact. PCA has a property of ranking the basis vectors. The one that has the highest eigenvalue will correspond to the principal component which has the highest variance and which explains the largest part of the data and etc. ICA does not have such property. In fact, the algorithm finds a basis up to a scaling factor which means that a basis vector might be randomly multiplied by -1. ICA implies some assumptions on the distribution of data: all sources, except at most 1, should have non-gaussian distribution and thus it can be viewed as a generalization of PCA to non-gaussian data. It turns out that PCA is a part of ICA. It is used to whiten the data matrix before doing ICA which means that, since PCA works with first two moments of distribution, PCA and ICA do not intersect at all. Although, as mentioned by Aires, Chédin, and Nadal (2000), PCA might provide a good starting point for iterative algorithm of ICA. The main difference between them is that they are designed to achieve different goals: PCA is designed to maximize variance with certain ”bonuses” as orthogonality, linear independence and dimensionality reduction. On the other hand, ICA is used for separation of components (Jolliffe (2002)), extending this separation to a strict sense of independence than just uncorrelatedness, and it might fail to return reasonable outcome if there is no mixture (linear on non-linear) of independent sources in the data. In contrast to factor analysis, ICA does not try to explain correlations between factors. Rather it assumes that ecologically valid factors will be independent. Both techniques are not robust to noise which is the main weakness when they are applied to financial data. Although several variations exist (see for example Ikeda and Toyama (2000), Voss, Rademacher, and Belkin (2013), Bingham and Hyvärinen (2000), Candès, Li, Ma, and Wright (2011)) and PCA and ICA might actually be used for noise separation. In particular (based on their properties) ICA is best in non-Gaussian noise separation and PCA in Gaussian respectively.
منابع مشابه
Comparison of MLP NN Approach with PCA and ICA for Extraction of Hidden Regulatory Signals in Biological Networks
The biologists now face with the masses of high dimensional datasets generated from various high-throughput technologies, which are outputs of complex inter-connected biological networks at different levels driven by a number of hidden regulatory signals. So far, many computational and statistical methods such as PCA and ICA have been employed for computing low-dimensional or hidden represe...
متن کاملComparison of multi-subject ICA methods for analysis of fMRI data.
Spatial independent component analysis (ICA) applied to functional magnetic resonance imaging (fMRI) data identifies functionally connected networks by estimating spatially independent patterns from their linearly mixed fMRI signals. Several multi-subject ICA approaches estimating subject-specific time courses (TCs) and spatial maps (SMs) have been developed, however, there has not yet been a f...
متن کاملICA vs. PCA Active Appearance Models: Application to Cardiac MR Segmentation
Statistical shape models generally use Principal Component Analysis (PCA) to describe the main directions of shape variation in a training set of example shapes. However, PCA has the restriction that the input data must be drawn from a Gaussian distribution and is only able to describe global shape variations. In this paper we evaluate the use of an alternative shape decomposition, Independent ...
متن کاملTowards the Development of next Generation Remote Sensing Technology – Erdas Imagine Incorporates a Higher Order Feature Extraction Technoque Based on Ica
Analysis of multi/hyperspectral imagery necessitates a selection of an optimal subset of bands in order to avoid the Hughes phenomena and parameter estimation problems due to interband correlation. This can be achieved by employing feature extraction techniques for significant reduction of data dimensionality. From the perspective of statistical pattern recognition, feature extraction refers to...
متن کاملIndependent Component Analysis in Statistical Shape Models
Statistical shape models generally use Principal Component Analysis (PCA) to describe the main directions of shape variation in a training set of example shapes. However, PCA assumes a number of restrictions on the data that do not always hold. In this paper we explore the use of an alternative shape decomposition, Independent Component Analysis (ICA), which does not assume a Gaussian distribut...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- CoRR
دوره abs/1709.10222 شماره
صفحات -
تاریخ انتشار 2017